# Encoder-Decoder Architecture
Mrt5 Large
MrT5 is an efficient byte-level language model based on ByT5 improvements, reducing input sequence length by approximately 50% through dynamic token merging technology
Large Language Model
Transformers Supports Multiple Languages

M
stanfordnlp
33
2
Shuka 1
Shuka v1 is a language model natively supporting Indian language audio understanding, combining a self-developed audio encoder with the Llama3-8B-Instruct decoder, enabling zero-shot multilingual question-answering tasks.
Audio-to-Text
Transformers Supports Multiple Languages

S
sarvamai
729
54
Pile T5 Base
Pile-T5 Base is an encoder-decoder model trained on The Pile dataset using the T5x library, trained for 2 million steps with MLM objective, approximately 2 trillion tokens.
Large Language Model
Transformers English

P
EleutherAI
50
19
Pile T5 Xxl
Pile-T5 XXL is an encoder-decoder model trained on The Pile dataset using the T5x library, employing a MLM objective similar to the original T5 model, trained for 2 million steps (approximately 2 trillion tokens).
Large Language Model
Transformers English

P
EleutherAI
44
28
Medicap
Apache-2.0
MedICap is an encoder-decoder model for medical image captioning, which won the championship in the ImageCLEFmedical Caption 2023 challenge.
Image-to-Text
Transformers

M
aehrc
475
2
Vlt5 Base Keywords
An encoder-decoder keyword generation model based on Google's Transformer architecture, supporting Polish and English, primarily used for extracting keywords from scientific paper abstracts.
Text Generation
Transformers Supports Multiple Languages

V
Voicelab
6,736
55
Bert Mini2bert Mini Finetuned Cnn Daily Mail Summarization
Apache-2.0
This is an encoder-decoder model based on the BERT-mini architecture, specifically fine-tuned for the CNN/Dailymail dataset for text summarization tasks.
Text Generation
Transformers English

B
mrm8488
140
5
Wav2vec2 Large Xlsr 53 German Gpt2
Apache-2.0
This is an automatic speech recognition encoder-decoder model trained on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 German dataset, combining the strengths of Wav2Vec2 and GPT2 architectures.
Speech Recognition
Transformers German

W
jsnfly
28
2
Bert2bert Turkish Paraphrase Generation
A Turkish paraphrase generation model based on the Bert2Bert architecture, used to generate sentences with the same meaning but different expressions.
Text Generation
Transformers Other

B
ahmetbagci
118
12
Roberta2roberta L 24 Cnn Daily Mail
Apache-2.0
An encoder-decoder model initialized with RoBERTa-Large, specifically designed for summarization tasks and fine-tuned on the CNN/DailyMail dataset.
Text Generation
Transformers English

R
google
128
6
Encoder Decoder Es
An encoder-decoder model fine-tuned on the cc_news_es_titles dataset for Spanish text processing tasks
Large Language Model
Transformers

E
amazon-sagemaker-community
121
0
Featured Recommended AI Models